Accessibility settings

Published on in Vol 28 (2026)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/91659, first published .
Deep Learning Algorithms Versus Radiologists in Digital Breast Tomosynthesis for Breast Cancer Detection: Systematic Review and Meta-Analysis

Deep Learning Algorithms Versus Radiologists in Digital Breast Tomosynthesis for Breast Cancer Detection: Systematic Review and Meta-Analysis

Deep Learning Algorithms Versus Radiologists in Digital Breast Tomosynthesis for Breast Cancer Detection: Systematic Review and Meta-Analysis

Authors of this article:

Shewen Lyu1 Author Orcid Image ;   Zepeng Wang1 Author Orcid Image ;   Yujing Mu1 Author Orcid Image ;   Luyao Wang1 Author Orcid Image ;   Xiaohua Pei1 Author Orcid Image

Beijing University of Chinese Medicine Third Affiliated Hospital, 51 Xiaoguan Street, Andingmenwai, Chaoyang District, Beijing, China

Corresponding Author:

Xiaohua Pei, MD


Background: Deep learning (DL) algorithms for digital breast tomosynthesis (DBT) have proliferated, demonstrating emerging potential in enhancing lesion detection and classification.

Objective: This study aimed to compare the diagnostic performance of DL algorithms for DBT with that of radiologists of varying experience and assess the clinical impact of DL assistance.

Methods: A systematic search of PubMed, Embase, Web of Science, and the Cochrane Library was conducted up to November 8, 2025. Included studies compared the performance of stand-alone DL algorithms for DBT, radiologist interpretation alone, and DL-assisted diagnosis. Study quality was assessed using the Prediction Model Risk of Bias Assessment Tool+Artificial Intelligence (PROBAST+AI). Performance metrics were pooled using bivariate random effects and generalized linear mixed models.

Results: A total of 13 studies with 38,565 patients were included in the final analysis. Stand-alone DL algorithms achieved a pooled sensitivity of 0.88 (95% CI 0.80-0.93), specificity of 0.74 (95% CI 0.59-0.85), and area under the receiver operating characteristic curve (AUC) of 0.89 (95% CI 0.86-0.92). While DL performance showed no statistically significant difference compared to all radiologists (AUC=0.89 vs 0.88; P=.64) or senior radiologists (AUC=0.89 vs 0.90; P=.48), DL demonstrated significantly superior sensitivity compared to junior radiologists (0.88 vs 0.76; P=.03). Notably, DL assistance did not statistically improve diagnostic metrics for radiologists across any experience level. Meta-regression identified validation methods as a significant source of heterogeneity.

Conclusions: DL algorithms for DBT exhibited strong diagnostic proficiency and showed higher sensitivity than junior radiologists, suggesting their potential utility as adjunctive tools to help reduce oversight in less experienced settings. However, given that DL assistance did not significantly elevate overall human performance, current models act primarily as supplementary aids rather than definitive clinical tools. Future prospective multimodal studies are warranted to validate these findings and optimize clinical integration.

Trial Registration: PROSPERO CRD420251242858; https://www.crd.york.ac.uk/PROSPERO/view/CRD420251242858

J Med Internet Res 2026;28:e91659

doi:10.2196/91659

Keywords



Breast cancer is the most commonly diagnosed cancer and the leading cause of cancer deaths among women worldwide, with an estimated 2.3 million new cases and 666,000 deaths occurring worldwide in 2022. This profound global burden underscores the urgency for early and accurate detection to improve prognosis and reduce the burden of invasive treatments [1,2]. While imaging serves as the cornerstone of screening, the transition from digital mammography to digital breast tomosynthesis (DBT) has revolutionized clinical practice. By providing quasi-3D volumetric data, DBT significantly alleviates the issue of tissue superposition, thereby increasing cancer detection rates and reducing unnecessary recall rates [3].

Despite these advantages, DBT introduces new challenges. The substantial increase in image volume significantly prolongs interpretation time, contributing to radiologist fatigue and potential cognitive overload [4,5]. Furthermore, despite overall improvements in detection rates, challenges regarding false positives (FPs) and false negatives (FNs) persist, with specific subtypes such as invasive lobular carcinoma prone to being overlooked [4,6]. The interpretation of DBT imagery is also heavily dependent on the radiologist’s experience, which increases subjectivity and the risk of misdiagnosis [7].

To address these diagnostic bottlenecks, deep learning (DL) algorithms for DBT have been increasingly developed, demonstrating emerging potential in enhancing lesion detection and classification. These algorithms aim to assist radiologists by extracting complex feature representations that may be imperceptible to the human eye [7]. However, results across the literature are inconsistent. While some pivotal studies suggest DL superiority [8], others indicate that algorithms may struggle with FPs or lack generalizability across different vendors and populations [9,10]. Moreover, the comparative performance of DL against radiologists of varying expertise (eg, junior and senior radiologists) remains a subject of ongoing debate [9,10]. This review includes data from 38,565 patients across the included studies.

Given the rapid accumulation of new evidence and the heterogeneity of study designs, a rigorous synthesis of current data is warranted. Therefore, the purpose of this systematic review and meta-analysis was to comprehensively compare the relative diagnostic performance and added value of DL algorithms vs radiologists of varying experience levels and evaluate potential factors influencing the diagnostic performance of these algorithms.


The meta-analysis was carried out in full compliance with the PRISMA-DTA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Diagnostic Test Accuracy) guidelines [11]. Furthermore, the protocol for this study was registered with PROSPERO (CRD420251242858).

Ethical Considerations

This was a systematic review and meta-analysis, so ethics approval and consent to participate are not applicable.

Search Strategy

A comprehensive literature search was conducted up to November 8, 2025, across 4 electronic databases: PubMed, Embase, Web of Science, and Cochrane Library. The search strategy, designed and executed to maximize sensitivity, used a combination of free-text terms and controlled vocabulary (eg, MeSH [Medical Subject Headings] terms in PubMed). Key concepts included three domains: (1) artificial intelligence (AI; eg, “Deep Learning”), (2) the target disease (eg, “Breast Neoplasms”), and (3) the imaging modality (eg, “Digital Breast Tomosynthesis”). No restrictions were placed on language, publication date, or study type. Two independent reviewers (SL and YM) performed initial title and abstract screening followed by full-text assessment of potentially eligible studies. To ensure literature saturation, the reference lists of all included articles were manually screened. The full, detailed search syntax for each database is provided in Table S1 in Multimedia Appendix 1.

Inclusion and Exclusion Criteria

Studies were selected according to the participants, index test, target condition, reference standard, outcomes, and setting framework: women undergoing breast cancer screening via DBT (participants); evaluated the performance of stand-alone DL algorithms for DBT, independent radiologist interpretation, or DL-assisted radiologist diagnosis (index test); breast cancer, confirmed via histopathology for positive cases (target condition); final diagnosis based on histopathology (for positive cases) or clinical imaging follow-up (for negative cases; reference standard); primary outcomes, including diagnostic performance measures (sensitivity, specificity, and area under the receiver operating characteristic [ROC] curve [AUC]), and secondary outcomes, comprising clinical impact metrics (detection rate, positive predictive value [PPV], and recall rate; outcomes); and original studies using retrospective or prospective cohorts from screening programs or clinical databases (setting).

In addition, we systematically excluded studies whose titles and abstracts were clearly irrelevant, as well as noneligible publication types, including reviews, case reports, conference abstracts, meta-analyses, and letters to the editor. Furthermore, studies that did not involve DL for DBT; without algorithm comparison; and with true positive (TP), FP, true negative (TN), and FN data not available were also excluded. The screening process was conducted in duplicate by 2 independent reviewers (SL and ZW), with any disagreements resolved through consultation with a third reviewer (XP).

Quality Assessment and Certainty of Evidence

We used the updated Prediction Model Risk of Bias Assessment Tool+AI (PROBAST+AI) quality assessment tool [12], which replaces the original 2019 Prediction Model Risk of Bias Assessment Tool instrument. This tool adopts a 2-phase structure comprising model development and model evaluation. Each phase includes 7 domains addressing participants, data sources, predictors, outcome assessment, and analytical approaches. For each domain, the risk-of-bias judgment is categorized as low, high, or unclear based on responses to predefined signaling questions. The full set of signaling questions and evaluation templates can be found in Tables S2 and S3 in Multimedia Appendix 1. To ensure objectivity and accuracy in the evaluation process, 2 reviewers (SL and LW) independently assessed the risk of bias in the included studies. The certainty of evidence was evaluated using the Grading of Recommendations Assessment, Development, and Evaluation (GRADE) framework [13]. Detailed evaluation items, decision rules, and domain-specific judgments can be found in Table S4 in Multimedia Appendix 1.

Data Extraction

Two reviewers (SL and YM) independently extracted data from the included full-text articles, and disagreements were resolved through consultation with a third reviewer (XP). Extracted items included patient characteristics, details of DL methods, DBT imaging parameters, stand-alone DL performance, radiologist performance, and radiologist performance with DL assistance. Because most studies did not report full contingency tables, we used two strategies to derive TP, FP, FN, and TN values: (1) back calculation using reported sensitivity, specificity, number of positive cases, and total sample size; and (2) redigitizing ROC curves using the GetData software to obtain optimal sensitivity and specificity based on the Youden index. For DL algorithms, only validation set results were collected. When multiple DL models or radiologists were reported with overlapping cohorts, only the best-performing result (highest AUC) was extracted to avoid duplication.

Outcome Measures

The primary outcome measures included the sensitivity, specificity, and AUC of DL algorithms, radiologists, and radiologists assisted by DL, as well as detection rate, PPV, and recall rate. Secondary outcomes focused on diagnostic performance stratified by radiologist experience both with and without DL assistance. Sensitivity (TP rate) reflected the ability to correctly identify cancer cases and was calculated as TP/(TP + FN) × 100%. Specificity (TN rate) represented the ability to correctly identify noncancer cases, calculated as TN/(TN + FP) × 100%. AUC summarizes overall discriminative ability. Detection rate was defined as TP/N × 100%, representing the proportion of cancers correctly detected in the screening population. PPV was calculated as TP/(TP + FP) × 100%, and recall rate was calculated as (TP + FP)/N × 100%. Radiologists with less than 5 years of experience were classified as junior, those with 5 or more years of experience were classified as senior, and studies without explicit experience data were categorized as unspecified.

Statistical Analysis

Given the expected methodological heterogeneity across studies, we applied a bivariate random-effects model [14] to pool sensitivity, specificity, and AUC estimates. For PPV, recall rate, and detection rate, log-transformation was performed prior to synthesis using a random-effects generalized linear mixed model framework. Differences in pooled diagnostic performance were assessed using a mean Z-test, with statistical significance defined as a P value of less than .05. Heterogeneity was quantified using the Higgins I2 statistic [15]. For substantial heterogeneity (I2>50%), bivariate box plots were used to explore potential sources, and multivariable meta-regression was conducted for DL algorithms to evaluate the impact of validation strategy, study design, region of interest, and data splitting method. Temporal changes in DL performance were examined using bubble plots, whereas violin plots were used to visualize differences in radiologist performance before and after DL assistance. Fagan nomograms were generated to assess the clinical implications for patients. The assessment of publication bias was conducted using the Deeks funnel plot [16]. Analyses were executed using Stata (version 15.1; StataCorp) with the midas and metadta commands, as well as R (version 4.5.1; R Foundation for Statistical Computing) using the ggplot2 and tidyverse packages.


Literature Search and Study Selection

A total of 1076 potentially relevant records were identified through the initial database search. Of these 1076 records, after removing 422 (39.2%) duplicates, 654 (60.8%) proceeded to title and abstract screening. During this stage, of the 654 remaining articles, 634 (96.9%) were excluded due to clear irrelevance or noneligible publication types. The remaining 20 articles were assessed in full text. Following detailed evaluation, of these 20 articles, 2 (10%) [17,18] were excluded because the data required to construct contingency tables (TP, FP, TN, and FN) were unavailable, 4 (20%) [19-22] were excluded because they were not research on DL for DBT, and 3 (15%) [23-25] lacked direct algorithm comparisons. In addition, 2 eligible records were identified from nondatabase sources. Ultimately, 13 studies met all the inclusion criteria and were incorporated into the meta-analysis [8-10,26-35]. The study selection process followed the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines and is illustrated in Figure 1.

Figure 1. PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flow diagram illustrating the study selection process. DBT: digital breast tomosynthesis; DL: deep learning; FN: false negative; FP: false positive; TN: true negative; TP: true positive.

Study Description and Quality Assessment

A total of 13 studies met the eligibility criteria. Among these, 46.2% (n=6) [8,9,28,31,32,35] reported internal validation cohorts (n=25,885 patients), and 61.5% (n=8) [9,10,26,27,29,30,33,34] reported external validation cohorts (n=12,680 patients; n=1, 7.7% of the studies reported both). In total, 61.5% (n=8) of the studies [9,10,26,27,29,30,33,34] evaluated stand-alone DL algorithms, and 84.6% (n=11) [8-10,26,28-33,35] assessed changes in radiologist performance before and after DL assistance. All studies (n=13, 100%) [8-10,26-35] reported outcomes for the overall radiologist group, whereas 15.4% (n=2) [10,35] included junior radiologists, 30.8% (n=4) [10,27,31,35] included senior radiologists, and 69.2% (n=9) [8,9,26,28-30,32-34] involved radiologists with unspecified experience levels. The included studies were published between 2017 and 2025. A total of 92.3% (n=12) [8-10,27-35] of the studies were retrospective in design, and 7.7% (n=1) [26] were prospective; all used pathological biopsy as the reference standard. Detailed characteristics are summarized in Tables S5 to S11 in Multimedia Appendix 1.

The quality assessment using the PROBAST+AI tool is shown in Figure 2 and Tables S2 and S3 in Multimedia Appendix 1. For model development, no study was rated as high risk in terms of either overall risk of bias or applicability concerns. For model evaluation, 50% (4/8) of the studies were judged to have high overall risk of bias, whereas none showed high applicability concerns. Overall, although the risk of bias in the evaluation phase was notable, the applicability of the included studies was generally acceptable. According to the GRADE framework, the certainty of the evidence ranged from low to moderate, primarily downgraded due to risk of bias and imprecision, as detailed in Table S4 in Multimedia Appendix 1.

Figure 2. Risk of bias and applicability concerns of the included studies using the Prediction Model Risk of Bias Assessment Tool+Artificial Intelligence (PROBAST+AI) tool: (A) summary of PROBAST+AI assessment for model development and (B) summary of PROBAST+AI assessment for model evaluation.

DL Algorithms

The sensitivity of DL algorithms was found to be 0.88 (95% CI 0.80-0.93; I2=96.22%; low certainty), and the specificity was 0.74 (95% CI 0.59-0.85; I2=99.63%; low certainty). The AUC was 0.89 (95% CI 0.86-0.92; low certainty), whereas the detection rate was 0.14 (95% CI 0.06-0.29; I2=98.8%; low certainty). The PPV was 0.41 (95% CI 0.18-0.70; I2=99.1%; low certainty), and the recall rate was 0.39 (95% CI 0.27-0.53; I2=99.7%; low certainty). As shown in Tables 1 and 2 and Figures S1 to S5 in Multimedia Appendix 1, the bubble chart indicates that the AUC values remained relatively stable from 2021 to 2024, as shown in Figure 3A.

Table 1. Deep learning (DL) algorithms for digital breast tomosynthesis (DBT) vs radiologists of different experience levels in terms of diagnostic performance outcomes for breast cancer diagnosis.
SubgroupSample size, nValidation datasets, nSensitivity (95% CI)Difference in sensitivityaSpecificity (95% CI)Difference in specificityaAUCb (95% CI)Difference in AUCa
Z-scoreP valueZ-scoreP valueZ-scoreP value
Overall DL for DBT12,55580.88 (0.80-0.93)c0.74 (0.59‐0.85)0.89 (0.86-0.92)
DL for DBT internal validation518210.96 (0.90-1.00)2.63.0090.40 (0.03-0.76)1.95.050.84 (0.78-0.89)1.62.11
DL for DBT external validation737370.85 (0.77-0.90)0.78 (0.65-0.87)0.89 (0.86-0.91)
DL for DBT vs all radiologists1.33.181.51.130.46.64
DL for DBT12,55580.88 (0.80-0.93)0.74 (0.59-0.85)0.89 (0.86-0.92)
All radiologists36,245140.83 (0.79-0.86)0.86 (0.75-0.92)0.88 (0.85-0.91)
DL for DBT vs junior radiologists2.12.031.35.18
DL for DBT12,55580.88 (0.80-0.93)0.74 (0.59-0.85)0.89 (0.86-0.92)
Junior radiologists31020.76 (0.66-0.84)0.94 (0.48-1.00)
DL for DBT vs senior radiologists0.50.621.28.20.50.62
DL for DBT12,55580.88 (0.80-0.93)0.74 (0.59-0.85)0.89 (0.86-0.92)
Senior radiologists96140.86 (0.81-0.90)0.93 (0.48-1.00)0.90 (0.87-0.92)

aOn the basis of 2-sided Z-test.

bAUC: area under the receiver operating characteristic curve.

cNot applicable.

Table 2. Deep learning (DL) for digital breast tomosynthesis (DBT) vs radiologists of different experience levels in terms of clinical impact outcomes for breast cancer diagnosis.
SubgroupSample size nValidation datasets, nDetection rate (95% CI)Difference in detection rateaPPVb (95% CI)Difference in PPVaRecall rate (95% CI)Difference in recall ratea
Z-scoreP valueZ-scoreP valueZ-scoreP value
Overall DL for DBT12,55580.14 (0.06-0.29)c0.41 (0.18-0.70)0.39 (0.27-0.53)
DL for DBT internal validation518210.09 (0.08-0.09)0.87.380.13 (0.12-0.15)2.33.020.63 (0.62-0.65)4.04<.001
DL for DBT external validation737370.15 (0.06-0.33)0.47 (0.19-0.76)0.36 (0.24-0.50)
DL for DBT vs all radiologists0.18.911.27.210.96.34
DL for DBT12,55580.14 (0.06-0.29)0.41 (0.18-0.70)0.39 (0.27-0.53)
All radiologists36,245140.15 (0.06-0.30)0.62 (0.40-0.79)0.30 (0.19-0.45)
DL for DBT vs junior radiologists5.07>.993.41<.0012.05.04
DL for DBT12,55580.14 (0.06-0.29)0.41 (0.18-0.70)0.39 (0.27-0.53)
Junior radiologists31020.47 (0.42-0.53)0.87 (0.81-0.91)0.55 (0.47-0.63)
DL for DBT vs senior radiologists1.94.053.28.0010.25.81
DL for DBT12,55580.14 (0.06-0.29)0.41 (0.18-0.70)0.39 (0.27-0.53)
Senior radiologists96140.36 (0.19-0.57)0.90 (0.70-0.97)0.43 (0.22-0.68)

aOn the basis of 2-sided Z-test.

bPPV: positive predictive value.

cNot applicable.

Figure 3. Bubble plot and violin plots: (A) bubble plot of temporal trends in area under the receiver operating characteristic curve (AUC) performance of different deep learning (DL) models, (B) violin plots of diagnostic outcomes for all radiologists before and after DL assistance, (C) violin plots of diagnostic outcomes for junior radiologists before and after DL assistance, and (D) violin plots of diagnostic outcomes for senior radiologists before and after DL assistance. AI: artificial intelligence; CNN: convolutional neural network; NA: not available; PPV: positive predictive value; ResNet: residual neural network.

All Radiologists

The sensitivity of all radiologists was found to be 0.83 (95% CI 0.79-0.86; I2=70.78%; low certainty), and the specificity was 0.86 (95% CI 0.75-0.92; I2=99.35%; low certainty). The AUC was 0.88 (95% CI 0.85-0.91; low certainty), whereas the detection rate was 0.15 (95% CI 0.06-0.30; I2=99.4%; moderate certainty). The PPV was 0.62 (95% CI 0.40-0.79; I2=98%; low certainty), and the recall rate was 0.30 (95% CI 0.19-0.45; I2=99.5%; low certainty). As shown in Table 1 and 2 and Figures S6 to S10 in Multimedia Appendix 1, there were no statistically significant differences between the performance of DL algorithms and that of all radiologists across all metrics.

Radiologists With Different Levels of Experience

DL algorithms for DBT achieved markedly higher sensitivity than junior radiologists (0.88 vs 0.76; Z=2.12; P=.03). However, DL algorithms for DBT exhibited significantly lower PPV than junior radiologists (0.41 vs 0.87; Z=3.41; P<.001) and significantly lower recall rate than junior radiologists (0.39 vs 0.55; Z=2.05; P=.04). Additionally, DL algorithms for DBT had a significantly lower PPV than senior radiologists (0.41 vs 0.90; Z=3.28; P=.001). These findings are illustrated in Tables 1 and 2.

Changes Before and After DL Assistance

With DL assistance, there were no significant improvements in any of the outcome measures for all radiologists. Similarly, there were no significant improvements in any of the outcome measures for junior or senior radiologists. These findings are shown in Tables S12 and S13 in Multimedia Appendix 1.

Heterogeneity Testing: Bivariate Box Plots and Meta-Regression

The bivariate box plots suggested that the studies by Shoshan et al [9] and Pinto et al [29] might have contributed to the heterogeneity of DL algorithms, whereas those by Bassi et al [10] and Resch et al [27] might have been sources of heterogeneity among all radiologists, as shown in Figures 4A and 4B [9,10,27,29,35]. Meta-regression indicated that this heterogeneity primarily arose from differences in validation methods (internal validation vs external validation; specificity P=.05) and geographic regions (Europe vs North America; sensitivity P<.001), as shown in Table S14 in Multimedia Appendix 1.

Figure 4. Bivariate box plots and Deeks funnel plots: (A) bivariate box plot for stand-alone deep learning (DL) for digital breast tomosynthesis (DBT) reading, (B) bivariate box plot for radiologist reading, (C) Deeks funnel plot for stand-alone DL for DBT reading, and (D) Deeks funnel plot for radiologist reading [9,10,27,29,35]. ESS: effective sample size.

Sensitivity Analysis

To assess the robustness of the primary findings, 2 sensitivity analyses were conducted. First, after excluding outliers identified by the bivariate box plot graphical diagnostic method, the analysis yielded a sensitivity of 0.86 (95% CI 0.78‐0.91), specificity of 0.74 (95% CI 0.62‐0.83), AUC of 0.88 (95% CI 0.85‐0.91), detection rate of 0.13 (95% CI 0.05‐0.33), PPV of 0.38 (95% CI 0.14‐0.69), and recall of 0.36 (95% CI 0.23‐0.53). Second, after excluding studies assessed as having a high risk of bias using the PROBAST+AI tool in the validation set, the recalculated pooled effect sizes were as follows: sensitivity of 0.86 (95% CI 0.77‐0.91), specificity of 0.80 (95% CI 0.57‐0.93), AUC of 0.90 (95% CI 0.87‐0.92), detection rate of 0.21 (95% CI 0.15‐0.27), PPV of 0.61 (95% CI 0.37‐0.81), and recall of 0.40 (95% CI 0.26‐0.56). These results were consistent with those of the primary analysis, indicating that the overall conclusions regarding diagnostic performance were robust, neither unduly influenced by high-risk studies nor biased by individual outliers. These findings are shown in Table S15 in Multimedia Appendix 1.

Clinical Application Value and Publication Bias

The Deeks funnel plot asymmetry test showed no evidence of publication bias (0.98 vs 0.09), as illustrated in Figures 4C and 4D [9,10,27,29,35]. Using the median prevalence from the included studies as the prior probability, the Fagan nomogram for DL algorithms (median prevalence 26%, IQR 13.84%–38.95%) indicated a positive posttest probability of 54%, while that for all radiologists (median prevalence 26%, IQR 12.60%–55.01%) indicated a positive posttest probability of 67%, as shown in Figure 5.

Figure 5. Fagan nomograms illustrating the posttest probability of breast cancer with digital breast tomosynthesis (DBT) reading: (A) stand-alone deep learning (DL) for DBT reading and (B) radiologist reading without DL assistance. LR: likelihood ratio.

Principal Findings

Our systematic review and meta-analysis demonstrated that DL algorithms using DBT achieved diagnostic proficiency comparable to the aggregate performance of radiologists and senior experts while exhibiting significantly superior sensitivity compared to junior radiologists. The observed parity between DL and senior radiologists suggests that current computational models have attained a level of pattern recognition equivalent to seasoned clinical judgment. However, the fact that DL has not significantly superseded senior radiologists implies that, while algorithms facilitate standardization, they currently lack the nuance required to outperform complex human decision-making [36-38]. Conversely, the significant performance gap regarding junior practitioners likely stems from the algorithms’ robust capacity to systematically analyze vast volumes of annotated data, enabling the detection of subtle or occult lesions that may be obscured by the subjectivity or limited pattern recognition characteristic of early-career experience [39]. By minimizing the specific oversight errors associated with inexperience [40], DL models effectively function as a “safety net” enhancing diagnostic sensitivity. Consequently, these findings support the integration of DL not as a stand-alone replacement for expert review but as a vital adjunctive tool to augment the capabilities of less experienced radiologists and bridge diagnostic disparities in resource-constrained environments.

Regarding clinical impact metrics, our initial analysis showed that DL algorithms had lower PPV and recall rates than junior radiologists. However, these findings should be interpreted with caution as they likely reflect spectrum bias and the limited number of studies available for the junior subgroup (2/13, 15.4%). Specifically, the junior radiologist data were derived from highly enriched cohorts, with a malignancy prevalence of 70% in the study by Chae et al [35] and 57% positive findings in the study by Bassi et al [10]. As PPV is a prevalence-dependent metric, the exceptionally high disease burden in these datasets artificially inflated the pooled estimates for the junior subgroup, reflecting fundamental differences in the underlying test populations rather than intrinsic radiologist superiority.

More importantly, an essential clinical counterpoint to the sensitivity advantage of DL algorithms is the significantly lower pooled PPV of stand-alone DL algorithms (0.41) compared to all radiologists (0.62) and senior radiologists (0.90). A PPV of 0.41 implies that less than half of the cases flagged by the DL algorithms are truly malignant, corresponding to a substantial FP burden. This finding is particularly contradictory as one of the primary, well-documented advantages of DBT over traditional mammography is its ability to reduce unnecessary recall rates. Using low-PPV DL algorithms as a primary or concurrent reading tool could erode this advantage, leading to a cascade of adverse consequences. These include unnecessary biopsies, heightened patient psychological distress and anxiety, and the overconsumption of health care resources—costs that are thoroughly documented in breast screening literature. Therefore, we urge a cautious approach to claims advocating for current DL algorithms as universal standardization tools. Their integration should be highly context dependent: in settings experiencing a severe shortage of experienced radiologists, the sensitivity benefits may offset the FP costs; however, in fully resourced screening environments with available senior radiologists, the net clinical benefit of stand-alone DL remains uncertain. Before recommending these models as reliable standardization tools, future deployment frameworks must incorporate well-defined, acceptable safety thresholds for FP rates and prospectively evaluate downstream clinical and economic outcomes.

Interestingly, our analysis comparing radiologist performance with and without DL assistance revealed no statistically significant incremental benefit. This lack of synergistic enhancement likely results from a “ceiling effect,” where experienced radiologists and high-performing algorithms achieve similar diagnostic performance, leaving limited room for improvement [41,42]. Furthermore, the clinical utility of DL assistants is constrained by “automation bias” and the opaque nature of “black box” algorithms [7,43]. When radiologist confidence in the algorithm is low—or when the AI serves merely as a concurrent reader without explainable features—the translation of algorithmic output into improved decision-making is diminished [44]. This suggests that, for DL to provide substantive added value, future systems must move beyond binary classification to provide interpretability and context-aware insights.

In comparison with a previous meta-analysis published by Yoon et al [45] in 2023, which reported that stand-alone AI significantly outperformed radiologists (AUC=0.90 vs 0.79) for breast cancer diagnosis, our study presents more conservative findings (AUC=0.89 vs 0.88; no significant difference). This discrepancy is attributable to the substantial expansion of the evidence base in our review. By incorporating approximately 3 times the number of studies, our analysis mitigates small-study effects that may have overestimated AI superiority in previous reviews. Our analysis suggests that the diagnostic performance of DL algorithms is currently comparable with the aggregate performance of the overall radiologist population in DBT interpretation but not demonstrably superior.

Beyond prior analyses, we assessed the specific impact of the DL assistant by comparing diagnostic performance without and with the DL assistant, offering critical insights into the human-AI interaction. We also incorporated essential clinical impact metrics such as detection rate, PPV, and recall rate to assess practical utility beyond discrimination accuracy. Furthermore, the methodological rigor was strengthened through the application of the PROBAST+AI tool for quality assessment and the GRADE approach to evaluate the certainty of evidence. These enhancements and stratified analyses collectively supply new, high-quality evidence that clarifies the current application of DL in breast imaging.

The extreme heterogeneity observed in several pooled metrics (I2>98%) warrants careful interpretation. Although meta-regression identified the validation strategy (internal vs external) as a significant contributor to heterogeneity in specificity (P=.05), this single factor cannot account for the near-complete variance in metrics such as PPV and recall rate. We posit that the residual heterogeneity reflects fundamental differences inherent to the DL algorithms themselves. First, algorithm design thresholds vary considerably across vendors: some systems are deliberately calibrated toward high sensitivity for triage purposes [27], whereas others prioritize specificity to minimize FP recalls, resulting in divergent operating points along the ROC curve. Second, training set composition, including differences in malignancy prevalence, patient demographics, geographic regions, mammographic density distribution, and scanner hardware, introduces domain-specific biases that cannot be homogenized through statistical pooling alone [28]. Third, variability in FP tolerance thresholds across health care systems and screening programs further compounds interstudy differences [9]. Given this landscape, future studies should prioritize standardized reporting of operating thresholds and algorithm configurations to facilitate more meaningful comparisons.

Additionally, bivariate box plots pinpointed the studies by Shoshan et al [9] and Pinto et al [29] as distinct outliers. This heterogeneity is mechanistically explainable: Shoshan et al [9] optimized their algorithm for a “triage” workflow, deliberately sacrificing specificity to maximize sensitivity for ruling out normal cases, resulting in a skewed performance profile compared to standard diagnostic models. Meanwhile, the divergence in the study by Pinto et al [29] is attributable to a small sample size (N=190), which introduces substantial statistical instability and random variation into the results.

Looking ahead, most of the DL algorithms included in our study were confined to single-modality DBT data, lacking the contextual depth provided by supplementary imaging and clinical history. Future iterations must prioritize multimodal integration to synthesize comprehensive diagnostic insights by correlating findings with the patient’s clinical background [46]. Implementation is further constrained by technical and systemic barriers, including data scarcity, regulatory challenges, and limited generalizability. While emerging techniques such as few-shot learning and self-supervised models may address these gaps, sustained multidisciplinary efforts are essential to optimize AI safety and deliver comprehensive solutions that genuinely augment radiological practice [47]. Currently, the opacity of decision-making processes remains a critical hurdle, highlighting the urgent need for future AI systems to prioritize transparency and interpretability [48].

Some limitations of this meta-analysis should be considered when interpreting the results. First, the preponderance of retrospective designs introduces potential selection bias, necessitating validation through large-scale prospective trials. Second, to distinguish between independent datasets and mitigate the risk of patient overlap, we extracted data exclusively from the highest-performing DL algorithm in each study context. This methodological necessity may introduce an optimism bias, potentially leading to an overestimation of the average algorithmic performance. Third, the limited number of studies stratified by radiologist experience (particularly for juniors) constrains the statistical robustness of subgroup analyses. Fourth, because most included studies did not report complete contingency tables, we used the GetData software to redigitize published ROC curves and derived operating points using the Youden index. This approach introduces two layers of methodological limitation: (1) the manual redigitization process could introduce subjective measurement error, potentially introducing small but nonnegligible inaccuracies in the extracted sensitivity and specificity values; and (2) more critically, the Youden index–derived operating point represents a theoretical optimum rather than the actual clinical threshold used in each study’s real-world setting. Clinically deployed algorithms may operate at thresholds deliberately chosen to balance sensitivity and specificity according to institutional recall policies or regulatory requirements. Consequently, our extracted performance values may not faithfully reflect the true clinical performance of these algorithms as implemented, and this methodological limitation should be considered when interpreting the pooled estimates.

Conclusions

In conclusion, DL algorithms for DBT demonstrated strong diagnostic performance. Although initial subgroup analyses indicated potentially higher sensitivity than junior radiologists, these findings are based on limited studies and require extensive validation to confirm their reliability. The current lack of a significant incremental benefit in human-AI collaborative workflows suggests that AI implementation should be approached cautiously—not as a stand-alone replacement or an automatic performance booster but rather as a supplementary “second opinion.” Future research must prioritize the development of explainable AI and prospective multimodal studies to better define the true synergistic potential of human-machine collaboration in breast cancer screening.

Acknowledgments

During the preparation of this work, the authors used DeepSeek-V3.2 to improve readability and language quality. After using this tool, the authors reviewed and edited the content as needed and take full responsibility for the content of the publication.

Funding

This research was funded by the National Natural Science Foundation of China (grant 8237153504).

Data Availability

The datasets generated or analyzed during this study are available from the corresponding author on reasonable request.

Authors' Contributions

Conceptualization: SL

Data curation: XP

Formal analysis: SL, ZW, YM, LW, XP

Investigation: ZW, YM, LW, XP

Methodology: ZW, YM, LW, XP

Software: SL

Writing—original draft: SL

Writing—review and editing: LW, XP

All authors read and approved the final manuscript.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Supplementary tables and figures for the systematic review and meta-analysis.

DOCX File, 3202 KB

Checklist 1

PRISMA checklist.

DOCX File, 20 KB

  1. Giaquinto AN, Sung H, Miller KD, et al. Breast cancer statistics, 2022. CA Cancer J Clin. Nov 2022;72(6):524-541. [CrossRef] [Medline]
  2. Bray F, Laversanne M, Sung H, et al. Global cancer statistics 2022: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. 2024;74(3):229-263. [CrossRef] [Medline]
  3. Chong A, Weinstein SP, McDonald ES, Conant EF. Digital breast tomosynthesis: concepts and clinical practice. Radiology. Jul 2019;292(1):1-14. [CrossRef] [Medline]
  4. Gao Y, Moy L, Heller SL. Digital breast tomosynthesis: update on technology, evidence, and clinical practice. Radiographics. 2021;41(2):321-337. [CrossRef] [Medline]
  5. Raichand S, Blaya-Novakova V, Berber S, Livingstone A, Noguchi N, Houssami N. Digital breast tomosynthesis for breast cancer diagnosis in women with dense breasts and additional breast cancer risk factors: a systematic review. Breast. Oct 2024;77:103767. [CrossRef] [Medline]
  6. Wang R, Chen F, Chen H, et al. Deep learning in digital breast tomosynthesis: current status, challenges, and future trends. MedComm (2020). Jun 2025;6(6):e70247. [CrossRef] [Medline]
  7. Bai J, Posner R, Wang T, Yang C, Nabavi S. Applying deep learning in digital breast tomosynthesis for automatic breast cancer detection: a review. Med Image Anal. Jul 2021;71:102049. [CrossRef] [Medline]
  8. Balleyguier C, Arfi-Rouche J, Levy L, et al. Improving digital breast tomosynthesis reading time: a pilot multi-reader, multi-case study using concurrent Computer-Aided Detection (CAD). Eur J Radiol. Dec 2017;97:83-89. [CrossRef] [Medline]
  9. Shoshan Y, Bakalo R, Gilboa-Solomon F, et al. Artificial intelligence for reducing workload in breast cancer screening with digital breast tomosynthesis. Radiology. Apr 2022;303(1):69-77. [CrossRef] [Medline]
  10. Bassi E, Russo A, Oliboni E, et al. The role of an artificial intelligence software in clinical senology: a mammography multi-reader study. Radiol Med. Feb 2024;129(2):202-210. [CrossRef] [Medline]
  11. Salameh JP, Bossuyt PM, McGrath TA, et al. Preferred reporting items for systematic review and meta-analysis of diagnostic test accuracy studies (PRISMA-DTA): explanation, elaboration, and checklist. BMJ. Aug 14, 2020;370:m2632. [CrossRef] [Medline]
  12. Moons KG, Damen JA, Kaul T, et al. PROBAST+AI: an updated quality, risk of bias, and applicability assessment tool for prediction models using regression or artificial intelligence methods. BMJ. Mar 24, 2025;388:e082505. [CrossRef] [Medline]
  13. Gopalakrishna G, Mustafa RA, Davenport C, et al. Applying Grading of Recommendations Assessment, Development and Evaluation (GRADE) to diagnostic tests was challenging but doable. J Clin Epidemiol. Jul 2014;67(7):760-768. [CrossRef] [Medline]
  14. Arends LR, Hamza TH, van Houwelingen JC, Heijenbrok-Kal MH, Hunink MG, Stijnen T. Bivariate random effects meta-analysis of ROC curves. Med Decis Making. 2008;28(5):621-638. [CrossRef] [Medline]
  15. Higgins JP, Thompson SG. Quantifying heterogeneity in a meta-analysis. Stat Med. Jun 15, 2002;21(11):1539-1558. [CrossRef] [Medline]
  16. Deeks JJ, Macaskill P, Irwig L. The performance of tests of publication bias and other sample size effects in systematic reviews of diagnostic test accuracy was assessed. J Clin Epidemiol. Sep 2005;58(9):882-893. [CrossRef] [Medline]
  17. Morrell S, Hutel M, Lucena O, et al. Dynamic AI-assisted ipsilateral tissue matching for digital breast tomosynthesis. Eur J Radiol. Nov 2025;192:112356. [CrossRef] [Medline]
  18. Letter H, Peratikos M, Toledano A, et al. Use of artificial intelligence for digital breast tomosynthesis screening: a preliminary real-world experience. J Breast Imaging. May 22, 2023;5(3):258-266. [CrossRef] [Medline]
  19. Uematsu T, Nakashima K, Harada TL, Nasu H, Igarashi T. Artificial intelligence computer-aided detection enhances synthesized mammograms: comparison with original digital mammograms alone and in combination with tomosynthesis images in an experimental setting. Breast Cancer. Jan 2023;30(1):46-55. [CrossRef] [Medline]
  20. Dahlblom V, Andersson I, Lång K, Tingberg A, Zackrisson S, Dustler M. Artificial intelligence detection of missed cancers at digital mammography that were detected at digital breast tomosynthesis. Radiol Artif Intell. 2021;3(6):e200299. [CrossRef] [Medline]
  21. Uematsu T, Nakashima K, Harada TL, Nasu H, Igarashi T. Comparisons between artificial intelligence computer-aided detection synthesized mammograms and digital mammograms when used alone and in combination with tomosynthesis images in a virtual screening setting. Jpn J Radiol. Jan 2023;41(1):63-70. [CrossRef] [Medline]
  22. Larsen M, Aglen CF, Lee CI, et al. Artificial intelligence evaluation of 122 969 mammography examinations from a population-based screening program. Radiology. Jun 2022;303(3):502-511. [CrossRef] [Medline]
  23. Plimpton SR, Milch H, Sears C, et al. External validation of a commercial artificial intelligence algorithm on a diverse population for detection of false negative breast cancers. J Breast Imaging. Jan 25, 2025;7(1):16-26. [CrossRef] [Medline]
  24. Dahlblom V, Dustler M, Zackrisson S, Tingberg A. Workload reduction of digital breast tomosynthesis screening using artificial intelligence and synthetic mammography: a simulation study. J Med Imaging (Bellingham). Nov 2025;12(Suppl 2):S22005. [CrossRef] [Medline]
  25. Berti A, Scapicchio C, Iacconi C, et al. An explainable-by-design end-to-end AI framework based on prototypical part learning for lesion detection and classification in Digital Breast Tomosynthesis images. Comput Struct Biotechnol J. 2025;27:2649-2660. [CrossRef] [Medline]
  26. van Winkel SL, Rodríguez-Ruiz A, Appelman L, et al. Impact of artificial intelligence support on accuracy and reading time in breast tomosynthesis image interpretation: a multi-reader multi-case study. Eur Radiol. Nov 2021;31(11):8682-8691. [CrossRef] [Medline]
  27. Resch D, Lo Gullo R, Teuwen J, et al. AI-enhanced mammography with digital breast tomosynthesis for breast cancer detection: clinical value and comparison with human performance. Radiol Imaging Cancer. Jul 2024;6(4):e230149. [CrossRef] [Medline]
  28. Raya-Povedano JL, Romero-Martín S, Elías-Cabot E, Gubern-Mérida A, Rodríguez-Ruiz A, Álvarez-Benito M. AI-based strategies to reduce workload in breast cancer screening with mammography and tomosynthesis: a retrospective evaluation. Radiology. Jul 2021;300(1):57-65. [CrossRef] [Medline]
  29. Pinto MC, Rodriguez-Ruiz A, Pedersen K, et al. Impact of artificial intelligence decision support using deep learning on breast cancer screening interpretation with single-view wide-angle digital breast tomosynthesis. Radiology. Sep 2021;300(3):529-536. [CrossRef] [Medline]
  30. Park EK, Kwak SY, Lee W, Choi JS, Kooi T, Kim EK. Impact of AI for digital breast tomosynthesis on breast cancer detection and interpretation time. Radiol Artif Intell. May 2024;6(3):e230318. [CrossRef] [Medline]
  31. Kim JG, Haslam B, Diab AR, et al. Impact of a categorical AI system for digital breast tomosynthesis on breast cancer interpretation by both general radiologists and breast imaging specialists. Radiol Artif Intell. Mar 2024;6(2):e230137. [CrossRef] [Medline]
  32. Elías-Cabot E, Romero-Martín S, Raya-Povedano JL, Brehl AK, Álvarez-Benito M. Impact of real-life use of artificial intelligence as support for human reading in a population-based breast cancer screening program with mammography and tomosynthesis. Eur Radiol. Jun 2024;34(6):3958-3966. [CrossRef] [Medline]
  33. Conant EF, Toledano AY, Periaswamy S, et al. Improving accuracy and efficiency with concurrent use of artificial intelligence for digital breast tomosynthesis. Radiol Artif Intell. Jul 31, 2019;1(4):e180096. [CrossRef] [Medline]
  34. Chen IE, Joines MM, Capiro N, et al. Commercial artificial intelligence versus radiologists: NPV and recall rate in large population-based digital mammography and tomosynthesis screening mammography cohorts. AJR Am J Roentgenol. Dec 2025;225(6):e2532889. [CrossRef] [Medline]
  35. Chae EY, Kim HH, Jeong JW, Chae SH, Lee S, Choi YW. Decrease in interpretation time for both novice and experienced readers using a concurrent computer-aided detection system for digital breast tomosynthesis. Eur Radiol. May 2019;29(5):2518-2525. [CrossRef] [Medline]
  36. Balkenende L, Teuwen J, Mann RM. Application of deep learning in breast cancer imaging. Semin Nucl Med. Sep 2022;52(5):584-596. [CrossRef] [Medline]
  37. Abdullah KA, Marziali S, Nanaa M, Escudero Sánchez L, Payne NR, Gilbert FJ. Deep learning-based breast cancer diagnosis in breast MRI: systematic review and meta-analysis. Eur Radiol. Aug 2025;35(8):4474-4489. [CrossRef] [Medline]
  38. Witowski J, Heacock L, Reig B, et al. Improving breast cancer diagnostics with deep learning for MRI. Sci Transl Med. Sep 28, 2022;14(664):eabo4802. [CrossRef] [Medline]
  39. Gu Y, Xu W, Lin B, et al. Deep learning based on ultrasound images assists breast lesion diagnosis in China: a multicenter diagnostic study. Insights Imaging. Jul 28, 2022;13(1):124. [CrossRef] [Medline]
  40. Lotter W, Diab AR, Haslam B, et al. Robust breast cancer detection in mammography and digital breast tomosynthesis using an annotation-efficient deep learning approach. Nat Med. Feb 2021;27(2):244-249. [CrossRef] [Medline]
  41. Xue P, Wang J, Qin D, et al. Deep learning in image-based breast and cervical cancer detection: a systematic review and meta-analysis. NPJ Digit Med. Feb 15, 2022;5(1):19. [CrossRef] [Medline]
  42. Tan H, Wu Q, Wu Y, et al. Mammography-based artificial intelligence for breast cancer detection, diagnosis, and BI-RADS categorization using multi-view and multi-level convolutional neural networks. Insights Imaging. May 21, 2025;16(1):109. [CrossRef] [Medline]
  43. Galati F, Maroncelli R, De Nardo C, et al. Deep learning with transfer learning on digital breast tomosynthesis: a radiomics-based model for predicting breast cancer risk. Diagnostics (Basel). Jun 26, 2025;15(13):1631. [CrossRef] [Medline]
  44. Dan Q, Xu Z, Burrows H, Bissram J, Stringer JS, Li Y. Diagnostic performance of deep learning in ultrasound diagnosis of breast cancer: a systematic review. NPJ Precis Oncol. Jan 27, 2024;8:21. [CrossRef]
  45. Yoon JH, Strand F, Baltzer PAT, et al. Standalone AI for breast cancer detection at screening digital mammography and digital breast tomosynthesis: a systematic review and meta-analysis. Radiology. Jun 2023;307(5):e222639. [CrossRef] [Medline]
  46. Rao VM, Hla M, Moor M, et al. Multimodal generative AI for medical image interpretation. Nature. Mar 2025;639(8056):888-896. [CrossRef] [Medline]
  47. Sosna J, Joskowicz L, Saban M. Navigating the AI landscape in medical imaging: a critical analysis of technologies, implementation, and implications. Radiology. Jun 2025;315(3):e240982. [CrossRef] [Medline]
  48. Champendal M, Müller H, Prior JO, Dos Reis CS. A scoping review of interpretability and explainability concerning artificial intelligence methods in medical imaging. Eur J Radiol. Dec 2023;169:111159. [CrossRef] [Medline]


AI: artificial intelligence
AUC: area under the receiver operating characteristic curve
DBT: digital breast tomosynthesis
DL: deep learning
FN: false negative
FP: false positive
GRADE: Grading of Recommendations Assessment, Development, and Evaluation
MeSH: Medical Subject Headings
PPV: positive predictive value
PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses
PRISMA-DTA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Diagnostic Test Accuracy
PROBAST+AI: Prediction Model Risk of Bias Assessment Tool+Artificial Intelligence
ROC: receiver operating characteristic
TN: true negative
TP: true positive


Edited by Matthew Balcarras; submitted 18.Jan.2026; peer-reviewed by Fateme Sheida, Zekai Yu; final revised version received 26.Mar.2026; accepted 26.Mar.2026; published 06.May.2026.

Copyright

© Shewen Lyu, Zepeng Wang, Yujing Mu, Luyao Wang, Xiaohua Pei. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 6.May.2026.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.